Manipulation of the prosodic features of vocal tract length, nasality and articulatory precision using articulatory synthesis
نویسندگان
چکیده
Vocal emotions, as well as different speaking styles and speaker traits are characterized by a complex interplay of multiple prosodic features. Natural sounding speech synthesis with the ability to control such paralinguistic aspects requires the manipulation of the corresponding prosodic features. With traditional concatenative speech synthesis it is easy to manipulate the “primary” prosodic features pitch, duration, and intensity, but it is very hard to individually control “secondary” prosodic features like phonation type, vocal tract length, articulatory precision and nasality. These secondary features can be controlled more directly with parametric synthesis methods. In the present study we analyze the ability of articulatory speech synthesis to control secondary prosodic features by rule. To this end, nine German words were resynthesized with the software VocalTractLab 2.1 and then manipulated in different ways at the articulatory level to vary vocal tract length, articulatory precision and degree of nasality. Listening tests showed that most of the intended prosodic manipulations could be reliably identified with recognition rates between 7796 %. Only the manipulations to increase articulatory precision were hardly recognized. The results suggest that rule-based manipulations in articulatory synthesis are generally sufficient for the convincing synthesis of secondary prosodic features at the word level.
منابع مشابه
Articulatory synthesis using a vocal-tract model of variable length
A method of articulatory synthesis using a vocal-tract model with variable length is proposed. The vocal-tract length is derived prior to the unique determination of vocal-tract area parameters incorporated with a codebook that maps formants to vocal-tract length is used. A two-dimensional interpolation function for irregularly spaced data is conducted to confine vocal-tract length in the first...
متن کاملClassification of depression state based on articulatory precision
Neurophysiological changes in the brain associated with major depression disorder can disrupt articulatory precision in speech production. Motivated by this observation, we address the hypothesis that articulatory features, as manifested through formant frequency tracks, can help in automatically classifying depression state. Specifically, we investigate the relative importance of vocal tract f...
متن کاملProsodic timing analysis for articulatory re-synthesis using a bank of resonators with an adaptive oscillator
A method for the analysis of prosodic-level temporal structure is introduced. The method is based on measured phase angles of an oscillator as that oscillator is made to synchronize with reference points in a signal. Reference points are the predicted peaks of acoustic change as determined by the output of a bank of tuned resonators. A framework for articulatory resynthesis is then described. J...
متن کاملA Kinematic Study of Prosodic Structure in Articulatory and Manual Gestures: Results from a Novel Method of Data Collection
The primary goal of this work is to examine prosodic structure as expressed concurrently through articulatory and manual gestures. Specifically, we investigated the effects of phrase-level prominence (Experiment 1) and of prosodic boundaries (Experiments 2 and 3) on the kinematic properties of oral constriction and manual gestures. The hypothesis guiding this work is that prosodic structure wil...
متن کاملArticulatory VCV Synthesis from EMA Data
This paper reports experiments in synthesizing VCV sequences with French unvoiced stop or fricative consonants, using a time-domain simulation of the vocal-tract system. The necessary dynamics of the vocal-tract shape are derived in two steps: first, time-varying parameters of an articulatory model are calculated automatically from electromagnetic articulography (EMA) data, using a method previ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- Computer Speech & Language
دوره 41 شماره
صفحات -
تاریخ انتشار 2017